Scalable Learning of Non-Decomposable Objectives

نویسندگان

  • Elad Eban
  • Mariano Schain
  • Alan Mackey
  • Ariel Gordon
  • Ryan Rifkin
  • Gal Elidan
چکیده

Modern retrieval systems are often driven by an underlying machine learning model. The goal of such systems is to identify and possibly rank the few most relevant items for a given query or context. Thus, such systems are typically evaluated using a ranking-based performance metric such as the area under the precision-recall curve, the Fβ score, precision at fixed recall, etc. Obviously, it is desirable to train such systems to optimize the metric of interest. In practice, due to the scalability limitations of existing approaches for optimizing such objectives, large-scale retrieval systems are instead trained to maximize classification accuracy, in the hope that performance as measured via the true objective will also be favorable. In this work we present a unified framework that, using straightforward building block bounds, allows for highly scalable optimization of a wide range of ranking-based objectives. We demonstrate the advantage of our approach on several real-life retrieval problems that are significantly larger than those considered in the literature, while achieving substantial improvement in performance over the accuracyobjective baseline. Proceedings of the 20 International Conference on Artificial Intelligence and Statistics (AISTATS) 2017, Fort Lauderdale, Florida, USA. JMLR: W&CP volume 54. Copyright 2017 by the author(s).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Online and Stochastic Gradient Methods for Non-decomposable Loss Functions

Modern applications in sensitive domains such as biometrics and medicine frequently require the use of non-decomposable loss functions such as precision@k, F-measure etc. Compared to point loss functions such as hinge-loss, these offer much more fine grained control over prediction, but at the same time present novel challenges in terms of algorithm design and analysis. In this work we initiate...

متن کامل

Large-scale Learning With Global Non-Decomposable Objectives

Modern retrieval systems are often driven by an underlying machine learning model. The goal of such systems is to identify and possibly rank the few most relevant items for a given query or context. Thus, such systems are typically evaluated using a ranking-based performance metric such as the area under the precision-recall curve, the Fβ score, precision at fixed recall, etc. Obviously, it is ...

متن کامل

On the decomposable numerical range of operators

 ‎Let $V$ be an $n$-dimensional complex inner product space‎. ‎Suppose‎ ‎$H$ is a subgroup of the symmetric group of degree $m$‎, ‎and‎ ‎$chi‎ :‎Hrightarrow mathbb{C} $ is an irreducible character (not‎ ‎necessarily linear)‎. ‎Denote by $V_{chi}(H)$ the symmetry class‎ ‎of tensors associated with $H$ and $chi$‎. ‎Let $K(T)in‎ (V_{chi}(H))$ be the operator induced by $Tin‎ ‎text{End}(V)$‎. ‎Th...

متن کامل

Learning Bayesian Network Structures with Non-Decomposable Scores

OF THE DISSERTATION Learning Bayesian Network Structures with Non-Decomposable Scores

متن کامل

Learning Bayesian networks with ancestral constraints

We consider the problem of learning Bayesian networks optimally, when subject to background knowledge in the form of ancestral constraints. Our approach is based on a recently proposed framework for optimal structure learning based on non-decomposable scores, which is general enough to accommodate ancestral constraints. The proposed framework exploits oracles for learning structures using decom...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017